Goto

Collaborating Authors

 Tripoli District


Computational Linguistics Meets Libyan Dialect: A Study on Dialect Identification

Essgaer, Mansour, Massud, Khamis, Mamlook, Rabia Al, Ghmaid, Najah

arXiv.org Artificial Intelligence

This study investigates logistic regression, linear support vector machine, multinomial Naive Bayes, and Bernoulli Naive Bayes for classifying Libyan dialect utterances gathered from Twitter. The dataset used is the QADI corpus, which consists of 540,000 sentences across 18 Arabic dialects. Preprocessing challenges include handling inconsistent orthographic variations and non-standard spellings typical of the Libyan dialect. The chi-square analysis revealed that certain features, such as email mentions and emotion indicators, were not significantly associated with dialect classification and were thus excluded from further analysis. Two main experiments were conducted: (1) evaluating the significance of meta-features extracted from the corpus using the chi-square test and (2) assessing classifier performance using different word and character n-gram representations. The classification experiments showed that Multinomial Naive Bayes (MNB) achieved the highest accuracy of 85.89% and an F1-score of 0.85741 when using a (1,2) word n-gram and (1,5) character n-gram representation. In contrast, Logistic Regression and Linear SVM exhibited slightly lower performance, with maximum accuracies of 84.41% and 84.73%, respectively. Additional evaluation metrics, including log loss, Cohen kappa, and Matthew correlation coefficient, further supported the effectiveness of MNB in this task. The results indicate that carefully selected n-gram representations and classification models play a crucial role in improving the accuracy of Libyan dialect identification. This study provides empirical benchmarks and insights for future research in Arabic dialect NLP applications.




NativQA Framework: Enabling LLMs with Native, Local, and Everyday Knowledge

Alam, Firoj, Hasan, Md Arid, Laskar, Sahinur Rahman, Kutlu, Mucahid, Darwish, Kareem, Chowdhury, Shammur Absar

arXiv.org Artificial Intelligence

The rapid advancement of large language models (LLMs) has raised concerns about cultural bias, fairness, and their applicability in diverse linguistic and underrepresented regional contexts. To enhance and benchmark the capabilities of LLMs, there is a need to develop large-scale resources focused on multilingual, local, and cultural contexts. In this study, we propose the NativQA framework, which can seamlessly construct large-scale, culturally and regionally aligned QA datasets in native languages. The framework utilizes user-defined seed queries and leverages search engines to collect location-specific, everyday information. It has been evaluated across 39 locations in 24 countries and in 7 languages -- ranging from extremely low-resource to high-resource languages -- resulting in over 300K Question-Answer (QA) pairs. The developed resources can be used for LLM benchmarking and further fine-tuning. The framework has been made publicly available for the community (https://gitlab.com/nativqa/nativqa-framework).


Real-time Bangla Sign Language Translator

Pranto, Rotan Hawlader, Siddique, Shahnewaz

arXiv.org Artificial Intelligence

The human body communicates through various meaningful gestures, with sign language using hands being a prominent example. Bangla Sign Language Translation (BSLT) aims to bridge communication gaps for the deaf and mute community. Our approach involves using Mediapipe Holistic to gather key points, LSTM architecture for data training, and Computer Vision for realtime sign language detection with an accuracy of 94%. Keywords=Recurrent Neural Network, LSTM, Computer Vision, Bangla font.


Predicting Confinement Effect of Carbon Fiber Reinforced Polymers on Strength of Concrete using Metaheuristics-based Artificial Neural Networks

Wahab, Sarmed, Suleiman, Mohamed, Shabbir, Faisal, Mahmoudabadi, Nasim Shakouri, Waqas, Sarmad, Herl, Nouman, Ahmad, Afaq

arXiv.org Artificial Intelligence

Keywords: carbon fiber reinforced polymer, concrete, confinement effect, strength, particle swarm optimization, grey wolf optimizer, bat algorithm Abstract This article deals with the study of predicting the confinement effect of carbon fiber reinforced polymers (CFRPs) on concrete cylinder strength using metaheuristics-based artificial neural networks. Three metaheuristic models are implemented including particle swarm optimization (PSO), grey wolf optimizer (GWO), and bat algorithm (BA). These algorithms are trained on the data using an objective function of mean square error and their predicted results are validated against the experimental studies and finite element analysis. The study shows that the hybrid model of PSO predicted the strength of CFRP-confined concrete cylinders with maximum accuracy of 99.13% and GWO predicted the results with an accuracy of 98.17%. The high accuracy of axial compressive strength predictions demonstrated that these prediction models are a reliable solution to the empirical methods. The prediction models are especially suitable for avoiding full-scale time-consuming experimental tests that make the process quick and economical. 1 Introduction Fiber-reinforced polymer is a composite material comprising fibers of either glass, aramid, or carbon and a polymer matrix. These fibers improve the properties of the polymer matrix mechanically including its stiffness and strength. The popularity of these composites has increased significantly in civil engineering due to their ability to strengthen concrete structural members. FRPs can be used either as a bar or plates embedded in concrete as an internal reinforcement and can be used as an external reinforcement by wrapping FRP sheets to existing structural members. The FRP bars have significantly higher strength than the steel reinforcement bars. They are highly durable and resistant to chemicals, corrosion (Cousin et al. 2019, Ananthkumar et al. 2020, Zhang et al. 2020), and radiation, their higher strength-to-weight ratio (Zhou et al. 2019) makes them ideal for structures that require high strength but need not be heavy. They can be molded into any required shape that provides higher design flexibility. Moreover, it has a lower environmental impact (Lee and Jain 2009), unlike concrete and timber.


Statistical exploration of the Manifold Hypothesis

Whiteley, Nick, Gray, Annie, Rubin-Delanchy, Patrick

arXiv.org Machine Learning

The Manifold Hypothesis is a widely accepted tenet of Machine Learning which asserts that nominally high-dimensional data are in fact concentrated near a low-dimensional manifold, embedded in high-dimensional space. This phenomenon is observed empirically in many real world situations, has led to development of a wide range of statistical methods in the last few decades, and has been suggested as a key factor in the success of modern AI technologies. We show that rich and sometimes intricate manifold structure in data can emerge from a generic and remarkably simple statistical model -- the Latent Metric Model -- via elementary concepts such as latent variables, correlation and stationarity. This establishes a general statistical explanation for why the Manifold Hypothesis seems to hold in so many situations. Informed by the Latent Metric Model we derive procedures to discover and interpret the geometry of high-dimensional data, and explore hypotheses about the data generating mechanism. These procedures operate under minimal assumptions and make use of well known, scaleable graph-analytic algorithms.


Exploring Memorization in Fine-tuned Language Models

Zeng, Shenglai, Li, Yaxin, Ren, Jie, Liu, Yiding, Xu, Han, He, Pengfei, Xing, Yue, Wang, Shuaiqiang, Tang, Jiliang, Yin, Dawei

arXiv.org Artificial Intelligence

LLMs have shown great capabilities in various tasks but also exhibited memorization of training data, thus raising tremendous privacy and copyright concerns. While prior work has studied memorization during pre-training, the exploration of memorization during fine-tuning is rather limited. Compared with pre-training, fine-tuning typically involves sensitive data and diverse objectives, thus may bring unique memorization behaviors and distinct privacy risks. In this work, we conduct the first comprehensive analysis to explore LMs' memorization during fine-tuning across tasks. Our studies with open-sourced and our own fine-tuned LMs across various tasks indicate that fine-tuned memorization presents a strong disparity among tasks. We provide an understanding of this task disparity via sparse coding theory and unveil a strong correlation between memorization and attention score distribution. By investigating its memorization behavior, multi-task fine-tuning paves a potential strategy to mitigate fine-tuned memorization.


Investigating Group Distributionally Robust Optimization for Deep Imbalanced Learning: A Case Study of Binary Tabular Data Classification

Mustapha, Ismail. B., Hasan, Shafaatunnur, Nabbus, Hatem S Y, Montaser, Mohamed Mostafa Ali, Olatunji, Sunday Olusanya, Shamsuddin, Siti Maryam

arXiv.org Artificial Intelligence

Oversampling and undersampling are two common data resampling approaches used in DNN. Owing to increased data availability, novel learning However, the susceptibility of the former to noise and architectures and accessibility to commodity computational overfitting due to added samples [23] as well as the hardware devices, deep neural networks (DNNs) have become characteristic loss of valuable information peculiar with the the de facto tool for a wide range of machine learning (ML) latter [3] remain major drawbacks of this category of tasks in recent times; leading to state-of-the-art performance in imbalance methods. On the other hand, the core idea behind several computer vision, natural language processing and the cost sensitive methods is to assign different speech recognition tasks. DNNs are characterized by several misclassification cost/weights to the training samples to scale layers of hidden units that enable learning of useful up/down the misclassification errors depending on the class representations of a given data for improved model they belong [17, 24]. While there are several implementations performance [1, 2]. This alleviates the need for domain experts of this method, the most commonly used cost sensitive and hand-engineered features, a common prerequisite for approach in imbalanced deep learning research is reweighting traditional ML methods.


ProSky: NEAT Meets NOMA-mmWave in the Sky of 6G

Benfaid, Ahmed, Adem, Nadia, Elmaghbub, Abdurrahman

arXiv.org Artificial Intelligence

Rendering to their abilities to provide ubiquitous connectivity, flexibly and cost effectively, unmanned aerial vehicles (UAVs) have been getting more and more research attention. To take the UAVs' performance to the next level, however, they need to be merged with some other technologies like non-orthogonal multiple access (NOMA) and millimeter wave (mmWave), which both promise high spectral efficiency (SE). As managing UAVs efficiently may not be possible using model-based techniques, another key innovative technology that UAVs will inevitably need to leverage is artificial intelligence (AI). Designing an AI-based technique that adaptively allocates radio resources and places UAVs in 3D space to meet certain communication objectives, however, is a tough row to hoe. In this paper, we propose a neuroevolution of augmenting topologies NEAT framework, referred to as ProSky, to manage NOMA-mmWave-UAV networks. ProSky exhibits a remarkable performance improvement over a model-based method. Moreover, ProSky learns 5.3 times faster than and outperforms, in both SE and energy efficiency EE while being reasonably fair, a deep reinforcement learning DRL based scheme. The ProSky source code is accessible to use here: https://github.com/Fouzibenfaid/ProSky